basic feature
LLMs4Synthesis: Leveraging Large Language Models for Scientific Synthesis
Giglou, Hamed Babaei, D'Souza, Jennifer, Auer, Sören
In response to the growing complexity and volume of scientific literature, this paper introduces the LLMs4Synthesis framework, designed to enhance the capabilities of Large Language Models (LLMs) in generating high-quality scientific syntheses. This framework addresses the need for rapid, coherent, and contextually rich integration of scientific insights, leveraging both open-source and proprietary LLMs. It also examines the effectiveness of LLMs in evaluating the integrity and reliability of these syntheses, alleviating inadequacies in current quantitative metrics. Our study contributes to this field by developing a novel methodology for processing scientific papers, defining new synthesis types, and establishing nine detailed quality criteria for evaluating syntheses. The integration of LLMs with reinforcement learning and AI feedback is proposed to optimize synthesis quality, ensuring alignment with established criteria. The LLMs4Synthesis framework and its components are made available, promising to enhance both the generation and evaluation processes in scientific research synthesis.
- Asia > China > Hong Kong (0.05)
- Europe > Germany > Lower Saxony > Hanover (0.04)
- South America > Paraguay > Asunción > Asunción (0.04)
- (7 more...)
From Basic to Extra Features: Hypergraph Transformer Pretrain-then-Finetuning for Balanced Clinical Predictions on EHR
Xu, Ran, Lu, Yiwen, Liu, Chang, Chen, Yong, Sun, Yan, Hu, Xiao, Ho, Joyce C, Yang, Carl
Electronic Health Records (EHRs) contain rich patient information and are crucial for clinical research and practice. In recent years, deep learning models have been applied to EHRs, but they often rely on massive features, which may not be readily available for all patients. We propose HTP-Star, which leverages hypergraph structures with a pretrain-then-finetune framework for modeling EHR data, enabling seamless integration of additional features. Additionally, we design two techniques, namely (1) Smoothness-inducing Regularization and (2) Group-balanced Reweighting, to enhance the model's robustness during fine-tuning. Through experiments conducted on two real EHR datasets, we demonstrate that HTP-Star consistently outperforms various baselines while striking a balance between patients with basic and extra features.
- North America > United States > Florida > Hillsborough County > University (0.05)
- North America > United States > Pennsylvania (0.04)
- Europe > United Kingdom (0.04)
- Asia > Middle East > Israel (0.04)
- Research Report > New Finding (1.00)
- Research Report > Experimental Study (1.00)
- Information Technology > Data Science (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (1.00)
Acer Swift laptops debut AMD Ryzen 8000 CPUs, AI apps
Acer is refreshing its Swift Edge 16 and Swift Go 14 laptops, showcasing one of the first instances of AMD's latest Ryzen 8000 mobile family of processors as well as a new AI app which Acer is preloading on the PCs. Acer will shift its updated Swift Edge 16 beginning in March, for 1,299 and up. Acer's Swift Go 14 with the latest Ryzen chips inside will cost 699 and up, and will ship in April. Acer's Swift Edge 16 shipped last year with the older Ryzen 7840U inside. PCWorld's Swift Edge 16 (2023) review assigned it four out of five stars, praising its inky black display, strong performance, and light weight, but frowning at its plasticky construction and keyboard.
Feature Programming for Multivariate Time Series Prediction
Reneau, Alex, Hu, Jerry Yao-Chieh, Xu, Chenwei, Li, Weijian, Gilani, Ammar, Liu, Han
We introduce the concept of programmable feature Our key motivation comes from a novel dynamical Ising-like engineering for time series modeling and propose model, the spin-gas Glauber dynamics, originated from a a feature programming framework. This newly debuted gas-like interaction that includes momentum framework generates large amounts of predictive and acceleration information. By using spin-gas Glauber features for noisy multivariate time series while dynamics as the fundamental model for time series generating allowing users to incorporate their inductive bias processes at the smallest time scale, we explore the with minimal effort. The key motivation of our potential of treating time series as the path-sum of infinitesimal framework is to view any multivariate time series increments generated by a series of Markovian coin as a cumulative sum of fine-grained trajectory tosses following the spin-gas Glauber dynamics. From such increments, with each increment governed by a a fine-grained perspective, a set of operators is motivated for novel spin-gas dynamical Ising model.
- Pacific Ocean > North Pacific Ocean > San Francisco Bay (0.04)
- North America > United States > Massachusetts > Middlesex County > Reading (0.04)
- North America > United States > Hawaii > Honolulu County > Honolulu (0.04)
- (2 more...)
Machine learning framework for end-to-end implementation of Incident duration prediction
Ajit, Smrithi, Mouli, Varsha R, Knickerbocker, Skylar, Wood, Jonathan S.
Traffic congestion caused by non-recurring incidents such as vehicle crashes and debris is a key issue for Traffic Management Centers (TMCs). Clearing incidents in a timely manner is essential for improving safety and reducing delays and emissions for the traveling public. However, TMCs and other responders face a challenge in predicting the duration of incidents (until the roadway is clear), making decisions of what resources to deploy difficult. To address this problem, this research developed an analytical framework and end-to-end machine-learning solution for predicting incident duration based on information available as soon as an incident report is received. Quality predictions of incident duration can help TMCs and other responders take a proactive approach in deploying responder services such as tow trucks, maintenance crews or activating alternative routes. The predictions use a combination of classification and regression machine learning modules. The performance of the developed solution has been evaluated based on the Mean Absolute Error (MAE), or deviation from the actual incident duration as well as Area Under the Curve (AUC) and Mean Absolute Percentage Error (MAPE). The results showed that the framework significantly improved incident duration prediction compared to methods from previous research.
- North America > United States > Iowa (0.05)
- North America > United States > Maryland (0.04)
- North America > United States > District of Columbia > Washington (0.04)
- Transportation > Ground > Road (1.00)
- Government > Regional Government > North America Government > United States Government (0.46)
- Information Technology > Artificial Intelligence > Machine Learning > Performance Analysis > Accuracy (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Learning Graphical Models > Directed Networks > Bayesian Learning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Ensemble Learning (0.95)
- (2 more...)
Classification of Multi-Spectral Pixels by the Binary Diamond Neural Network
Classification is a process by which an item is assigned to a class. Classification is widely used in the animal kingdom. Identifying an item as food is classification. Assigning words to objects, actions, feelings, and situations is classification. The purpose of this work is to introduce a new neural network, the Binary Diamond, which can be used as a general purpose classification tool.
Gravitational Models Explain Shifts on Human Visual Attention
Zanca, Dario, Gori, Marco, Melacci, Stefano, Rufa, Alessandra
Visual attention refers to the human brain's ability to select relevant sensory information for preferential processing, improving performance in visual and cognitive tasks. It proceeds in two phases. One in which visual feature maps are acquired and processed in parallel. Another where the information from these maps is merged in order to select a single location to be attended for further and more complex computations and reasoning. Its computational description is challenging, especially if the temporal dynamics of the process are taken into account. Numerous methods to estimate saliency have been proposed in the last three decades. They achieve almost perfect performance in estimating saliency at the pixel level, but the way they generate shifts in visual attention fully depends on winner-take-all (WTA) circuitry. WTA is implemented} by the biological hardware in order to select a location with maximum saliency, towards which to direct overt attention. In this paper we propose a gravitational model (GRAV) to describe the attentional shifts. Every single feature acts as an attractor and {the shifts are the result of the joint effects of the attractors. In the current framework, the assumption of a single, centralized saliency map is no longer necessary, though still plausible. Quantitative results on two large image datasets show that this model predicts shifts more accurately than winner-take-all.
- Europe > Italy (0.04)
- Europe > France > Provence-Alpes-Côte d'Azur (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- Europe > Netherlands > South Holland > Dordrecht (0.04)
John Pearce on LinkedIn: Neural Networks Application for Small-Scale Tasks
Learn how to use multilayer #NeuralNetworks to generate additional features in the context of small dimensional data when the number of basic features varies from one to two dozen. Learn how to use multilayer #NeuralNetworks to generate additional features in the context of small dimensional data when the number of basic features varies from one to two dozen.
Program synthesis performance constrained by non-linear spatial relations in Synthetic Visual Reasoning Test
Yihe, Lu, Lowe, Scott C., Lewis, Penelope A., van Rossum, Mark C. W.
Despite remarkable advances in automated visual recognition by machines, some visual tasks remain challenging for machines. Fleuret et al. (2011) introduced the Synthetic Visual Reasoning Test (SVRT) to highlight this point, which required classification of images consisting of randomly generated shapes based on hidden abstract rules using only a few examples. Ellis et al. (2015) demonstrated that a program synthesis approach could solve some of the SVRT problems with unsupervised, few-shot learning, whereas they remained challenging for several convolutional neural networks trained with thousands of examples. Here we re-considered the human and machine experiments, because they followed different protocols and yielded different statistics. We thus proposed a quantitative reintepretation of the data between the protocols, so that we could make fair comparison between human and machine performance. We improved the program synthesis classifier by correcting the image parsings, and compared the results to the performance of other machine agents and human subjects. We grouped the SVRT problems into different types by the two aspects of the core characteristics for classification: shape specification and location relation. We found that the program synthesis classifier could not solve problems involving shape distances, because it relied on symbolic computation which scales poorly with input dimension and adding distances into such computation would increase the dimension combinatorially with the number of shapes in an image. Therefore, although the program synthesis classifier is capable of abstract reasoning, its performance is highly constrained by the accessible information in image parsings.
- Europe > United Kingdom > England > Nottinghamshire > Nottingham (0.14)
- North America > United States > California > Santa Clara County > Palo Alto (0.04)
- North America > Canada > Ontario > Toronto (0.04)
- Europe > Slovenia > Drava > Municipality of Benedikt > Benedikt (0.04)
Natural representation of composite data with replicated autoencoders
Negri, Matteo, Bergamini, Davide, Baldassi, Carlo, Zecchina, Riccardo, Feinauer, Christoph
Generative processes in biology and other fields often produce data that can be regarded as resulting from a composition of basic features. Here we present an unsupervised method based on autoencoders for inferring these basic features of data. The main novelty in our approach is that the training is based on the optimization of the `local entropy' rather than the standard loss, resulting in a more robust inference, and enhancing the performance on this type of data considerably. Algorithmically, this is realized by training an interacting system of replicated autoencoders. We apply this method to synthetic and protein sequence data, and show that it is able to infer a hidden representation that correlates well with the underlying generative process, without requiring any prior knowledge.